Scalable Parallel Factorizations of SDD Matrices and Efficient Sampling for Gaussian Graphical Models
نویسندگان
چکیده
Motivated by a sampling problem basic to computational statistical inference, we develop a nearly optimal algorithm for a fundamental problem in spectral graph theory and numerical analysis. Given an n × n SDDM matrix M, and a constant −1 ≤ p ≤ 1, our algorithm gives efficient access to a sparse n× n linear operator C̃ such that M p ≈ C̃C̃ ⊤ . The solution is based on factoring M into a product of simple and sparse matrices using squaring and spectral sparsification. For M with m non-zero entries, our algorithm takes work nearly-linear in m, and polylogarithmic depth on a parallel machine with m processors. This gives the first sampling algorithm that only requires nearly linear work and n i.i.d. random univariate Gaussian samples to generate i.i.d. random samples for n-dimensional Gaussian random fields with SDDM precision matrices. For sampling this natural subclass of Gaussian random fields, it is optimal in the randomness and nearly optimal in the work and parallel complexity. In addition, our sampling algorithm can be directly extended to Gaussian random fields with SDD precision matrices.
منابع مشابه
Efficient Sampling for Gaussian Graphical Models via Spectral Sparsification
Motivated by a sampling problem basic to computational statistical inference, we develop a toolset based on spectral sparsification for a family of fundamental problems involving Gaussian sampling, matrix functionals, and reversible Markov chains. Drawing on the connection between Gaussian graphical models and the recent breakthroughs in spectral graph theory, we give the first nearly linear ti...
متن کاملScalable Log Determinants for Gaussian Process Kernel Learning
For applications as varied as Bayesian neural networks, determinantal point processes, elliptical graphical models, and kernel learning for Gaussian processes (GPs), one must compute a log determinant of an n× n positive definite matrix, and its derivatives – leading to prohibitive O(n) computations. We propose novel O(n) approaches to estimating these quantities from only fast matrix vector mu...
متن کاملInnovated Scalable Efficient Estimation in Ultra - Large Gaussian Graphical Models
Large-scale precision matrix estimation is of fundamental importance yet challenging in many contemporary applications for recovering Gaussian graphical models. In this paper, we suggest a new approach of innovated scalable efficient estimation (ISEE) for estimating large precision matrix. Motivated by the innovated transformation, we convert the original problem into that of large covariance m...
متن کاملSparsified Cholesky Solvers for SDD linear systems
We show that Laplacian and symmetric diagonally dominant (SDD) matrices can be well approximated by linear-sized sparse Cholesky factorizations. Specifically, n × n matrices of these types have constant-factor approximations of the form LL , where L is a lowertriangular matrix with O(n) non-zero entries. This factorization allows us to solve linear systems in such matrices in O(n) work and O(lo...
متن کاملTrek Separation for Gaussian Graphical Models
Gaussian graphical models are semi-algebraic subsets of the cone of positive definite covariance matrices. Submatrices with low rank correspond to generalizations of conditional independence constraints on collections of random variables. We give a precise graph-theoretic characterization of when submatrices of the covariance matrix have small rank for a general class of mixed graphs that inclu...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1410.5392 شماره
صفحات -
تاریخ انتشار 2014